27 research outputs found
Recommended from our members
Decoding-complexity-aware HEVC encoding using a complexity–rate–distortion model
The energy consumption of Consumer Electronic (CE) devices during media playback is inexorably linked to the computational complexity of decoding compressed video. Reducing a CE device's the energy consumption is therefore becoming ever more challenging with the increasing video resolutions and the complexity of the video coding algorithms. To this end, this paper proposes a framework that alters the video bit stream to reduce the decoding complexity and simultaneously limits the impact on the coding efficiency. In this context, this paper (i) first performs an analysis to determine the trade-off between the decoding complexity, video quality and bit rate with respect to a reference decoder implementation on a General Purpose Processor (GPP) architecture. Thereafter, (ii) a novel generic decoding complexity-aware video coding algorithm is proposed to generate decoding complexity-rate-distortion optimized High Efficiency Video Coding (HEVC) bit streams.
The experimental results reveal that the bit streams generated by the proposed algorithm achieve 29.43% and 13.22% decoding complexity reductions for a similar video quality with minimal coding efficiency impact compared to the state-of-the-art approaches when applied to the HM16.0 and openHEVC decoder implementations, respectively. In addition, analysis of the energy consumption behavior for the same scenarios reveal up to 20% energy consumption reductions while achieving a similar video quality to that of HM 16.0 encoded HEVC bit streams
Content-adaptive feature-based CU size prediction for fast low-delay video encoding in HEVC
Determining the best partitioning structure of a Coding Tree Unit (CTU) is one of the most time consuming operations in HEVC encoding. Specifically, it is the evaluation of the quadtree hierarchy using the Rate-Distortion (RD) optimization that has the most significant impact on the encoding time, especially in the cases of High Definition (HD) and Ultra High Definition (UHD) videos. In order to expedite the encoding for low delay applications, this paper proposes a Coding Unit (CU) size selection and encoding algorithm for inter-prediction in the HEVC. To this end, it describes (i) two CU classification models based on Inter N×N mode motion features and RD cost thresholds to predict the CU split decision, (ii) an online training scheme for dynamic content adaptation, (iii) a motion vector reuse mechanism to expedite the motion estimation process, and finally introduces (iv) a computational complexity to coding efficiency trade-off process to enable flexible control of the algorithm. The experimental results reveal that the proposed algorithm achieves a consistent average encoding time performance ranging from 55% - 58% and 57%-61% with average Bjøntegaard Delta Bit Rate (BDBR) increases of 1.93% –
2.26% and 2.14% – 2.33% compared to the HEVC 16.0 reference software for the low delay P and low delay B configurations, respectively, across a wide range of content types and bit rates
Optimized resource distribution for interactive TV applications
This paper proposes a novel resource optimization scheme for cloud-based interactive television applications that are increasingly believed to be the future of television broadcasting and media consumption, in general. The varying distribution of groups of users and the need for on-the-fly media processing inherent to this type of application necessitates a mechanism to efficiently allocate the resources at both a content and network level. A heuristic solution is proposed in order to (a) generate end-to-end delay bound multicast trees for individual groups of users and (b) co-locate multiple multicast trees, such that a minimum group quality metric can be satisfied. The performance of the proposed heuristic solution is evaluated in terms of the serving probability (i.e., the resource utilization efficiency) and execution time of the resource allocation decision making process. It is shown that improvements in the serving probability of up to 50%, in comparison with existing resource allocation schemes, and several orders of magnitude reduction of the execution time, in comparison to the linear programming approach to solving the optimization problem, can be achieved
Recommended from our members
End user video quality prediction and coding parameters selection at the encoder for robust HEVC video transmission
Along with the rapid increase in the availability for high quality video formats such as HD (High Definition), UHD (Ultra HD) and HDR (High Dynamic Range), a huge demand for data rates during their transmission has become inevitable. Consequently, the role of video compression techniques has become crucially important in the process of mitigating the data rate requirements. Even though the latest video codec HEVC (High Efficiency Video Coding) has succeeded in significantly reducing the data rate compared to its immediate predecessor H.264/AVC (Advanced Video Coding), the HEVC coded videos in the meantime have become even more vulnerable to network impairments. Therefore, it is equally important to assess the consumers’ perceived quality degradation prior to transmitting HEVC coded videos over an error prone network, and to include error resilient features so as to minimize the adverse effects those impairments. To this end, this paper proposes a probabilistic model which accurately predicts the overall distortion of the decoded video at the encoder followed by an accurate QP-λ relationship which can be used in the RDO (Rate Distortion Optimization) process. During the derivation process of the probabilistic model, the impacts from the motion vectors, the pixels in the reference frames and the clipping operations are accounted and consequently the model is capable of minimizing the prediction error as low as 3.11% whereas the state-of-the-art methods can’t reach below 20.08% under identical conditions. Furthermore, the enhanced RDO process has resulted in 21.41%- 43.59% improvement in the BD-rate compared to the state-of-the-art error resilient algorithms
HRTF aided broadband DOA estimation using two microphones
Two sensor broadband direction of arrival (DOA) estimation suffers from an inherent lack of dimensionality due to having just two sensors, yet humans and other animals are able to overcome this limitation using subtle variations introduced by the ears. Application of existing DOA estimation techniques to such systems becomes complicated due to the ill-behaved nature of the Head Related Transfer Function (HRTF). In this paper we present a subband signal extraction and focussing technique which retains the diversity information of the HRTF. We then develop a framework for combining these signals for subspace DOA estimation and investigate the constraints imposed on the single and multi-source DOA estimation problems. Finally, estimation performance is compared with existing techniques and we find performance has improved to be comparable to human localisation abilities. © 2012 IEEE
Novel head related transfer function model for sound source localisation
Human beings have a remarkable ability to determine the direction of arrival of a sound and to separate sounds of interest. Replicating this ability is a challenging problem in audio signal processing. In this paper we present a model for the head related transfer function (HRTF) developed with the localisation objective in mind. This is achieved by splitting the 3D localisation cues in terms of two functions which can be independently evaluated. We illustrate the theory for calculating these functions and validate the results against actual HRTF data. We find the model to be a close match for a significant number of potential source locations. © 2010 IEEE
Novel head related transfer function model for sound source localisation
Human beings have a remarkable ability to determine the direction of arrival of a sound and to separate sounds of interest. Replicating this ability is a challenging problem in audio signal processing. In this paper we present a model for the head related transfer function (HRTF) developed with the localisation objective in mind. This is achieved by splitting the 3D localisation cues in terms of two functions which can be independently evaluated. We illustrate the theory for calculating these functions and validate the results against actual HRTF data. We find the model to be a close match for a significant number of potential source locations. © 2010 IEEE